Anthropic
$21
Input tokens/M
$105
Output tokens/M
200
Context Length
Tencent
$1
$4
32
Deepseek
$12
128
Bytedance
$16
$3.5
Iflytek
$2
-
8
Alibaba
$0.75
Baidu
64
$0.5
224
$0.8
$0.2
Huawei
Openai
$7.7
$30.8
$1.5
$1.6
Nitral-AI
Captain-Eris_Violet is an advanced language model developed through multi-stage supervised fine-tuning, QLoRA adapters, and GRPO-optimized RLHF, suitable for role-playing and dialogue generation.
sfairXC
A reward model based on the Gemma-2-9B architecture, trained using RLHF workflow, suitable for dialogue and reasoning tasks.
0x7o
The Violet series language models are specifically trained for instruction following and maintaining Russian dialogues. The fourth generation has been optimized with RLHF, offering stronger response capabilities and richer information content.
TheBloke
StableVicuna-13B is a dialogue model fine-tuned via RLHF based on Vicuna-13B v0, using 4-bit GPTQ quantization format
CarperAI
StableVicuna-13B is a fine-tuned version of the Vicuna-13B v0 model, enhanced through Reinforcement Learning from Human Feedback (RLHF) and Proximal Policy Optimization (PPO) on various dialogue and instruction datasets.
zen-E
A dialogue generation model based on OPT-1.3b, optimized through RLHF training using the DeepSpeed-Chat framework